Classification of linearly nonseparable patterns by linear threshold elements
نویسندگان
چکیده
Learning and convergence properties of linear threshold elements or perceptrons are well understood for the case where the input vectors (or the training sets) to the perceptron are linearly separable. Little is known, however, about the behavior of the perceptron learning algorithm when the training sets are linearly nonseparable. We present the first known results on the structure of linearly nonseparable training sets and on the behavior of perceptrons when the set of input vectors is linearly nonseparable. More precisely, we show that using the well known perceptron learning algorithm, a linear threshold element can learn the input vectors that are provably learnable, and identify those vectors that cannot be learned without committing errors. We also show how a linear threshold element can be used to learn large linearly separable subsets of any given nonseparable training set. In order to develop our results, we first establish formal characterizations of linearly nonseparable training sets and define learnable structures for such patterns. We also prove computational complexity results for the related learning problems. Next, based on such characterizations, we show that a perceptron does the best one can expect for linearly nonseparable sets of input vectors and learns as much as is theoretically possible.
منابع مشابه
Motion Contrast Classification Is a Linearly Nonseparable Problem
Sensitivity to image motion contrast, that is, the relative motion between different parts of the visual field, is a common and computationally important property of many neurons in the visual pathways of vertebrates. Here we illustrate that, as a classification problem, motion contrast detection is linearly nonseparable. In order to do so, we prove a theorem stating a sufficient condition for ...
متن کاملA New Threshold Unit Learning Algorithm
A new algorithm for learning a threshold unit is proposed. The Barycentric Correction Procedure (BCP) is an eecient substitute for the Perceptron and its enhanced versions such as the Thermal Perceptron or the Pocket algorithm. Based on geometrical concepts, the BCP is much more eecient than the Perceptron for learning linearly separable mappings. To deal with linearly nonseparable mappings, ex...
متن کاملBackpropagation Separates Where Perceptrons Do1
Feedforward nets with sigmoidal activation functions are often designed by minimizing a cost criterion. It has been pointed out before that this technique may be outperformed by the classical perceptron learning rule, at least on some problems. In this paper, we show that no such pathologies can arise if the error criterion is of a threshold LMS type, i.e., is zero for values “beyond” the desir...
متن کاملLipschitz - free Banach spaces
We show that when a linear quotient map to a separable Banach space X has a Lipschitz right inverse, then it has a linear right inverse. If a separable space X embeds isometrically into a Banach space Y , then Y contains an isometric linear copy of X. This is false for every nonseparable weakly compactly generated Banach space X. Canonical examples of nonseparable Banach spaces which are Lipsch...
متن کاملIncreasing the accuracy of the classification of diabetic patients in terms of functional limitation using linear and nonlinear combinations of biomarkers: Ramp AUC method
The Area under the ROC Curve (AUC) is a common index for evaluating the ability of the biomarkers for classification. In practice, a single biomarker has limited classification ability, so to improve the classification performance, we are interested in combining biomarkers linearly and nonlinearly. In this study, while introducing various types of loss functions, the Ramp AUC method and some of...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- IEEE transactions on neural networks
دوره 6 2 شماره
صفحات -
تاریخ انتشار 1995